Note: When clicking on a Digital Object Identifier (DOI) number, you will be taken to an external site maintained by the publisher.
                                            Some full text articles may not yet be available without a charge during the embargo (administrative interval).
                                        
                                        
                                        
                                            
                                                
                                             What is a DOI Number?
                                        
                                    
                                
Some links on this page may take you to non-federal websites. Their policies may differ from this site.
- 
            Multi-agent reinforcement learning (MARL) lies at the heart of a plethora of applications involving the interaction of a group of agents in a shared unknown environment. A prominent framework for studying MARL is Markov games, with the goal of finding various notions of equilibria in a sample-efficient manner, such as the Nash equilibrium (NE) and the coarse correlated equilibrium (CCE). However, existing sample-efficient approaches either require tailored uncertainty estimation under function approximation, or careful coordination of the players. In this paper, we propose a novel model-based algorithm, called VMG, that incentivizes exploration via biasing the empirical estimate of the model parameters towards those with a higher collective best-response values of all the players when fixing the other players’ policies, thus encouraging the policy to deviate from its current equilibrium for more exploration. VMG is oblivious to different forms of function approximation, and permits simultaneous and uncoupled policy updates of all players. Theoretically, we also establish that VMG achieves a near-optimal regret for finding both the NEs of two-player zero-sum Markov games and CCEs of multi-player general-sum Markov games under linear function approximation in an online environment, which nearly match their counterparts with sophisticated uncertainty quantification.more » « lessFree, publicly-accessible full text available July 13, 2026
- 
            Free, publicly-accessible full text available April 24, 2026
- 
            Free, publicly-accessible full text available April 24, 2026
- 
            Free, publicly-accessible full text available April 24, 2026
- 
            Free, publicly-accessible full text available April 6, 2026
- 
            Free, publicly-accessible full text available April 24, 2026
- 
            Free, publicly-accessible full text available January 1, 2026
- 
            Free, publicly-accessible full text available January 1, 2026
- 
            Free, publicly-accessible full text available December 16, 2025
- 
            Free, publicly-accessible full text available January 1, 2026
 An official website of the United States government
An official website of the United States government 
				
			 
					 
					
